Conditional selective inference for robust regression and outlier detection using piecewise-linear homotopy continuation

نویسندگان

چکیده

In this paper, we consider conditional selective inference (SI) for a linear model estimated after outliers are removed from the data. To apply SI framework, it is necessary to characterize events of how robust method identifies outliers. Unfortunately, existing SIs cannot be directly applied our problem because they applicable case where selection can represented by or quadratic constraints. We propose popular regressions such as least-absolute-deviation regression and Huber introducing new computational using convex optimization technique called homotopy method. show that proposed wide class outlier detection methods has good empirical performance on both synthetic data real experiments.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Multiple Linear Regression Models in Outlier Detection

Identifying anomalous values in the realworld database is important both for improving the quality of original data and for reducing the impact of anomalous values in the process of knowledge discovery in databases. Such anomalous values give useful information to the data analyst in discovering useful patterns. Through isolation, these data may be separated and analyzed. The analysis of outlie...

متن کامل

Robust SiZer for Exploration of Regression Structures and Outlier Detection

The SiZer methodology proposed by Chaudhuri & Marron (1999) is a valuable tool for conducting exploratory data analysis. In this article a robust version of SiZer is developed for the regression setting. This robust SiZer is capable of producing SiZer maps with different degrees of robustness. By inspecting such SiZer maps, either as a series of plots or in the form of a movie, the structures h...

متن کامل

Outlier Detection by Boosting Regression Trees

A procedure for detecting outliers in regression problems is proposed. It is based on information provided by boosting regression trees. The key idea is to select the most frequently resampled observation along the boosting iterations and reiterate after removing it. The selection criterion is based on Tchebychev’s inequality applied to the maximum over the boosting iterations of ...

متن کامل

Outlier Detection Using Nonconvex Penalized Regression

This paper studies the outlier detection problem from the point of view of penalized regressions. Our regression model adds one mean shift parameter for each of the n data points. We then apply a regularization favoring a sparse vector of mean shift parameters. The usual L1 penalty yields a convex criterion, but we find that it fails to deliver a robust estimator. The L1 penalty corresponds to ...

متن کامل

Outlier Detection using Granular Box Regression Methods

Granular computing (GrC) is an emerging computing paradigm of information processing. It concerns the processing of complex information entities called information granules, which arise in the process of data abstraction and derivation of knowledge from information. Granular computing is more a theoretical perspective, it encourages an approach to data that recognizes and exploits the knowledge...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Annals of the Institute of Statistical Mathematics

سال: 2022

ISSN: ['1572-9052', '0020-3157']

DOI: https://doi.org/10.1007/s10463-022-00846-2